In statistics, normality tests are used to determine if a data set is well-modeled by a normal distribution and to compute how likely it is for a random variable underlying the data set to be normally distributed.
More precisely, the tests are a form of model selection, and can be interpreted several ways, depending on one's interpretations of probability:
A normality test is used to determine whether sample data has been drawn from a normally distributed population (within some tolerance). A number of statistical tests, such as the Student's t-test and the one-way and two-way analysis of variance (ANOVA), require a normally distributed sample population.
A graphical tool for assessing normality is the normal probability plot, a quantile-quantile plot (QQ plot) of the standardized data against the standard normal distribution. Here the correlation between the sample data and normal quantiles (a measure of the goodness of fit) measures how well the data are modeled by a normal distribution. For normal data the points plotted in the QQ plot should fall approximately on a straight line, indicating high positive correlation. These plots are easy to interpret and also have the benefit that outliers are easily identified.
This test is useful in cases where one faces kurtosis risk – where large deviations matter – and has the benefit that it is very easy to compute and to communicate: non-statisticians can easily grasp that 6 σ events are very rare in normal distributions.
Some published works recommend the Jarque–Bera test, but the test has weakness. In particular, the test has low power for distributions with short tails, especially for bimodal distributions. Some authors have declined to include its results in their studies because of its poor overall performance.
Historically, the third and fourth standardized moments (skewness and kurtosis) were some of the earliest tests for normality. The Lin–Mudholkar test specifically targets asymmetric alternatives. The Jarque–Bera test is itself derived from skewness and kurtosis estimates. Mardia's multivariate skewness and kurtosis tests generalize the moment tests to the multivariate case. Other early include the ratio of the mean absolute deviation to the standard deviation and of the range to the standard deviation.
More recent tests of normality include the energy testSzékely, G. J. and Rizzo, M. L. (2005) A new test for multivariate normality, Journal of Multivariate Analysis 93, 58–80. (Székely and Rizzo) and the tests based on the empirical characteristic function (ECF) (e.g. Epps and Pulley,Epps, T. W., and Pulley, L. B. (1983). A test for normality based on the empirical characteristic function. Biometrika 70, 723–726. Henze–Zirkler,Henze, N., and Zirkler, B. (1990). A class of invariant and consistent tests for multivariate normality. Communications in Statistics – Theory and Methods 19, 3595–3617. BHEP testHenze, N., and Wagner, T. (1997). A new approach to the BHEP tests for multivariate normality. Journal of Multivariate Analysis 62, 1–23.). The energy and the ECF tests are powerful tests that apply for testing univariate or multivariate normality and are statistically consistent against general alternatives.
The normal distribution has the highest entropy of any distribution for a given standard deviation. There are a number of normality tests based on this property, the first attributable to Vasicek.
Spiegelhalter suggests using a Bayes factor to compare normality with a different class of distributional alternatives.Spiegelhalter, D.J. (1980). An omnibus test for normality for small samples. Biometrika, 67, 493–496. This approach has been extended by Farrell and Rogers-Stewart.Farrell, P.J., Rogers-Stewart, K. (2006) "Comprehensive study of tests for normality and symmetry: extending the Spiegelhalter test". Journal of Statistical Computation and Simulation, 76(9), 803 – 816.
|
|